Active Learning for Effectively Fine-Tuning Transfer Learning to Downstream Task

نویسندگان

چکیده

Language model (LM) has become a common method of transfer learning in Natural Processing (NLP) tasks when working with small labeled datasets. An LM is pretrained using an easily available large unlabelled text corpus and fine-tuned the labelled data to apply target (i.e., downstream) task. As designed capture linguistic aspects semantics, it can be biased features. We argue that exposing during fine-tuning instances diverse semantic (e.g., topical, linguistic, relations) present dataset will improve its performance on underlying propose Mixed Aspect Sampling (MAS) framework sample different use ensemble classifier classification performance. Experimental results show MAS performs better than random sampling as well state-of-the-art active models abuse detection where hard collect for building accurate classifier.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Active Learning with Transfer Learning

In sentiment classification, unlabeled user reviews are often free to collect for new products, while sentiment labels are rare. In this case, active learning is often applied to build a high-quality classifier with as small amount of labeled instances as possible. However, when the labeled instances are insufficient, the performance of active learning is limited. In this paper, we aim at enhan...

متن کامل

Active Task Selection for Multi-Task Learning

In multi-task learning, a learner is given a collection of prediction tasks and needs to solve all of them. In contrast to previous work, which required that annotated training data is available for all tasks, we consider a new setting, in which for some tasks, potentially most of them, only unlabeled training data is provided. Consequently, to solve all tasks, information must be transferred b...

متن کامل

Active Learning

This article has no abstract.

متن کامل

Improving Machine Learning Ability with Fine-Tuning

Item Response Theory (IRT) allows for measuring ability of Machine Learning models as compared to a human population. However, it is difficult to create a large dataset to train the ability of deep neural network models (DNNs). We propose Crowd-Informed Fine-Tuning (CIFT) as a new training process, where a pre-trained model is fine-tuned with a specialized supplemental training set obtained via...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: ACM Transactions on Intelligent Systems and Technology

سال: 2021

ISSN: ['2157-6904', '2157-6912']

DOI: https://doi.org/10.1145/3446343